Add MobileNet v3 to MLTS Models created with cl/319020977 expectedTop1 values from TFLite_CPU Sample results: http://hackney.lon.corp.google.com:8000/result.html Bug: 131831554 Test: build_and_run_benchmark.sh Change-Id: Ibc38c01295d20e4e14471178c8d6a9fd3aace396 Merged-In: Ibc38c01295d20e4e14471178c8d6a9fd3aace396 (cherry picked from commit d1f639bac98f6a2b51e98efe0050f7f384f30731)
diff --git a/README.txt b/README.txt index 1072501..c3f24c0 100644 --- a/README.txt +++ b/README.txt
@@ -109,6 +109,19 @@ Apache License, Version 2.0 ------------------------------------------------------------------ +- mobilenet_v3-(small_224_0.75_float|small_224_1.0_float|small_224_1.0_uint8|small-minimalistic_224_1.0_float|large_224_0.75_float|large_224_1.0_float|large_224_1.0_uint8|large-minimalistic_224_1.0_float|large-minimalistic_224_1.0_uint8).tflite +MobileNet TensorFlow Lite models based on +"Searching for MobileNetV3" +https://arxiv.org/abs/1905.02244 +Apache License, Version 2.0 + +Downloaded from +https://storage.googleapis.com/mobilenet_v3/checkpoints/v3-${variant}.tgz +on Jun 30 2020. +See also: https://github.com/tensorflow/models/tree/master/research/slim/nets/mobilenet +Golden output generated with ToT TensorFlow (Linux, CPU). + +------------------------------------------------------------------ Input files: ------------------------------------------------------------------ - ssd_mobilenet_v1_coco_*/tarmac.input diff --git a/assets/mobilenet_v3-large-minimalistic_224_1.0_float.tflite b/assets/mobilenet_v3-large-minimalistic_224_1.0_float.tflite new file mode 100644 index 0000000..c53e102 --- /dev/null +++ b/assets/mobilenet_v3-large-minimalistic_224_1.0_float.tflite Binary files differ
diff --git a/assets/mobilenet_v3-large-minimalistic_224_1.0_uint8.tflite b/assets/mobilenet_v3-large-minimalistic_224_1.0_uint8.tflite new file mode 100644 index 0000000..e822f5c --- /dev/null +++ b/assets/mobilenet_v3-large-minimalistic_224_1.0_uint8.tflite Binary files differ
diff --git a/assets/mobilenet_v3-large_224_0.75_float.tflite b/assets/mobilenet_v3-large_224_0.75_float.tflite new file mode 100644 index 0000000..d14905d --- /dev/null +++ b/assets/mobilenet_v3-large_224_0.75_float.tflite Binary files differ
diff --git a/assets/mobilenet_v3-large_224_1.0_float.tflite b/assets/mobilenet_v3-large_224_1.0_float.tflite new file mode 100644 index 0000000..57be893 --- /dev/null +++ b/assets/mobilenet_v3-large_224_1.0_float.tflite Binary files differ
diff --git a/assets/mobilenet_v3-large_224_1.0_uint8.tflite b/assets/mobilenet_v3-large_224_1.0_uint8.tflite new file mode 100644 index 0000000..e4baecb --- /dev/null +++ b/assets/mobilenet_v3-large_224_1.0_uint8.tflite Binary files differ
diff --git a/assets/mobilenet_v3-small-minimalistic_224_1.0_float.tflite b/assets/mobilenet_v3-small-minimalistic_224_1.0_float.tflite new file mode 100644 index 0000000..cc2bc94 --- /dev/null +++ b/assets/mobilenet_v3-small-minimalistic_224_1.0_float.tflite Binary files differ
diff --git a/assets/mobilenet_v3-small_224_0.75_float.tflite b/assets/mobilenet_v3-small_224_0.75_float.tflite new file mode 100644 index 0000000..a05fe60 --- /dev/null +++ b/assets/mobilenet_v3-small_224_0.75_float.tflite Binary files differ
diff --git a/assets/mobilenet_v3-small_224_1.0_float.tflite b/assets/mobilenet_v3-small_224_1.0_float.tflite new file mode 100644 index 0000000..de0d80e --- /dev/null +++ b/assets/mobilenet_v3-small_224_1.0_float.tflite Binary files differ
diff --git a/assets/mobilenet_v3-small_224_1.0_uint8.tflite b/assets/mobilenet_v3-small_224_1.0_uint8.tflite new file mode 100644 index 0000000..33fe6cd --- /dev/null +++ b/assets/mobilenet_v3-small_224_1.0_uint8.tflite Binary files differ
diff --git a/assets/models_list/mobilenet_topk_aosp.json b/assets/models_list/mobilenet_topk_aosp.json index 60a031b..6198e4e 100644 --- a/assets/models_list/mobilenet_topk_aosp.json +++ b/assets/models_list/mobilenet_topk_aosp.json
@@ -230,6 +230,165 @@ "labels": "image_classification/labels.txt", "preprocessor": "Inception" } + }, + { + "name": "mobilenet_v3-small_224_0.75_float_topk_aosp", + "modelFile": "mobilenet_v3-small_224_0.75_float", + "baselineSec": 0.060, + "evaluator": { + "className": "TopK", + "expectedTop1": 0.60536915 + }, + "inputSize": [1, 224, 224, 3], + "dataSize": 4, + "dataset": { + "inputPath": "image_classification", + "groundTruth": "image_classification/ground_truth_labels.txt", + "labels": "image_classification/labels.txt", + "preprocessor": "Inception" + } + }, + { + "name": "mobilenet_v3-small_224_1.0_float_topk_aosp", + "modelFile": "mobilenet_v3-small_224_1.0_float", + "baselineSec": 0.060, + "evaluator": { + "className": "TopK", + "expectedTop1": 0.6107383 + }, + "inputSize": [1, 224, 224, 3], + "dataSize": 4, + "dataset": { + "inputPath": "image_classification", + "groundTruth": "image_classification/ground_truth_labels.txt", + "labels": "image_classification/labels.txt", + "preprocessor": "Inception" + } + }, + { + "name": "mobilenet_v3-small-minimalistic_224_1.0_float_topk_aosp", + "modelFile": "mobilenet_v3-small-minimalistic_224_1.0_float", + "baselineSec": 0.060, + "evaluator": { + "className": "TopK", + "expectedTop1": 0.5885906 + }, + "inputSize": [1, 224, 224, 3], + "dataSize": 4, + "dataset": { + "inputPath": "image_classification", + "groundTruth": "image_classification/ground_truth_labels.txt", + "labels": "image_classification/labels.txt", + "preprocessor": "Inception" + } + }, + { + "name": "mobilenet_v3-large_224_0.75_float_topk_aosp", + "modelFile": "mobilenet_v3-large_224_0.75_float", + "baselineSec": 0.060, + "evaluator": { + "className": "TopK", + "expectedTop1": 0.63892615 + }, + "inputSize": [1, 224, 224, 3], + "dataSize": 4, + "dataset": { + "inputPath": "image_classification", + "groundTruth": "image_classification/ground_truth_labels.txt", + "labels": "image_classification/labels.txt", + "preprocessor": "Inception" + } + }, + { + "name": "mobilenet_v3-large_224_1.0_float_topk_aosp", + "modelFile": "mobilenet_v3-large_224_1.0_float", + "baselineSec": 0.060, + "evaluator": { + "className": "TopK", + "expectedTop1": 0.6577181 + }, + "inputSize": [1, 224, 224, 3], + "dataSize": 4, + "dataset": { + "inputPath": "image_classification", + "groundTruth": "image_classification/ground_truth_labels.txt", + "labels": "image_classification/labels.txt", + "preprocessor": "Inception" + } + }, + { + "name": "mobilenet_v3-large-minimalistic_224_1.0_float_topk_aosp", + "modelFile": "mobilenet_v3-large-minimalistic_224_1.0_float", + "baselineSec": 0.060, + "evaluator": { + "className": "TopK", + "expectedTop1": 0.6375839 + }, + "inputSize": [1, 224, 224, 3], + "dataSize": 4, + "dataset": { + "inputPath": "image_classification", + "groundTruth": "image_classification/ground_truth_labels.txt", + "labels": "image_classification/labels.txt", + "preprocessor": "Inception" + } + }, + { + "name": "mobilenet_v3-small_224_1.0_uint8_topk_aosp", + "modelFile": "mobilenet_v3-small_224_1.0_uint8", + "baselineSec": 0.060, + "evaluator": { + "className": "TopK", + "expectedTop1": 0.56510067 + }, + "inputSize": [1, 224, 224, 3], + "dataSize": 1, + "inputScale": 0.007812, + "inputZeroPoint": 128.0, + "dataset": { + "inputPath": "image_classification", + "groundTruth": "image_classification/ground_truth_labels.txt", + "labels": "image_classification/labels.txt", + "preprocessor": "Inception" + } + }, + { + "name": "mobilenet_v3-large_224_1.0_uint8_topk_aosp", + "modelFile": "mobilenet_v3-large_224_1.0_uint8", + "baselineSec": 0.060, + "evaluator": { + "className": "TopK", + "expectedTop1": 0.61879194 + }, + "inputSize": [1, 224, 224, 3], + "dataSize": 1, + "inputScale": 0.007812, + "inputZeroPoint": 128.0, + "dataset": { + "inputPath": "image_classification", + "groundTruth": "image_classification/ground_truth_labels.txt", + "labels": "image_classification/labels.txt", + "preprocessor": "Inception" + } + }, + { + "name": "mobilenet_v3-large-minimalistic_224_1.0_uint8_topk_aosp", + "modelFile": "mobilenet_v3-large-minimalistic_224_1.0_uint8", + "baselineSec": 0.060, + "evaluator": { + "className": "TopK", + "expectedTop1": 0.59395975 + }, + "inputSize": [1, 224, 224, 3], + "dataSize": 1, + "inputScale": 0.007812, + "inputZeroPoint": 128.0, + "dataset": { + "inputPath": "image_classification", + "groundTruth": "image_classification/ground_truth_labels.txt", + "labels": "image_classification/labels.txt", + "preprocessor": "Inception" + } } ] }